Goto

Collaborating Authors

 probability distribution


Entropic Neural Optimal Transport via Diffusion Processes

Neural Information Processing Systems

We propose a novel neural algorithm for the fundamental problem of computing the entropic optimal transport (EOT) plan between continuous probability distributions which are accessible by samples. Our algorithm is based on the saddle point reformulation of the dynamic version of EOT which is known as the Schrödinger Bridge problem. In contrast to the prior methods for large-scale EOT, our algorithm is end-to-end and consists of a single learning step, has fast inference procedure, and allows handling small values of the entropy regularization coefficient which is of particular importance in some applied problems. Empirically, we show the performance of the method on several large-scale EOT tasks.




A Missing statements and proofs 521 A.1 Statements for Section 3.1

Neural Information Processing Systems

Let a two-player Markov game where both players affect the transition. As we have seen in Section 2.1, in the case of unilateral deviation from joint policy Let a (possibly correlated) joint policy ˆ σ . By Lemma A.1, we know that Where the equality holds due to the zero-sum property, (1). An approximate NE is an approximate global minimum. An approximate global minimum is an approximate NE.






A Graphical Terminology An arbitrary graph

Neural Information Processing Systems

We refer the readers to ( Peters et al., 2017) for more detailed graphical terminology. We base our proof mostly on ( Kirsch, 2019). The first statement follows directly from the first theorem in ( Haviland, 1936). Without loss of generality, we reorder the variables according to reversed topological ordering, i.e. a Follows directly from Lemma 1. Lemma 4. Recall condition 2) in Causal de Finetti states that 8 i, 8 n 2 N: X The first equality holds by well-defindedness. The fourth equality follow from well-definedness.